7,622 research outputs found

    The capacity and attractor basins of associative memory models

    Get PDF
    The original publication is available at www.springerlink.com . Copyright SpringerThe performance characteristics of five variants of the Hopfield network are examined. Two performance metrics are used: memory capacity, and a measure of the size of basins of attraction. We find that the posttraining adjustment of processor thresholds has, at best, little or no effect on performance, and at worst a significant negative effect. The adoption of a local learning rule can, however, give rise to significant performance gains.Peer reviewe

    An investigation into the performance and representation of a stochastic evolutionary neural tree

    Get PDF
    Copyright Springer.The Stochastic Competitive Evolutionary Neural Tree (SCENT) is a new unsupervised neural net that dynamically evolves a representational structure in response to its training data. Uniquely SCENT requires no initial parameter setting as it autonomously creates appropriate parameterisation at runtime. Pruning and convergence are stochastically controlled using locally calculated heuristics. A thorough investigation into the performance of SCENT is presented. The network is compared to other dynamic tree based models and to a high quality flat clusterer over a variety of data sets and runs

    Comparative performances of stochastic competitive evolutionary neural tree (SCENT) with neural classifiers

    Get PDF
    A stochastic competitive evolutionary neural tree (SCENT) is described and evaluated against the best neural classifiers with equivalent functionality, using a collection of data sets chosen to provide a variety of clustering scenarios. SCENT is firstly shown to produce flat classifications at least as well as the other two neural classifiers used. Moreover its variability in performance over the data sets is shown to be small. In addition SCENT also produces a tree that can show any hierarchical structure contained in the data. For two real world data sets the tree captures hierarchical features of the data.Peer reviewe

    Input window size and neural network predictors

    Get PDF
    Neural network approaches to time series prediction are briefly discussed, and the need to specify an appropriately sized input window identified. Relevant theoretical results from dynamic systems theory are briefly introduced, and heuristics for finding the correct embedding dimension, and hence window size, are discussed. The method is applied to two time series and the resulting generalisation performance of the trained feedforward neural network predictors is analysed. It is shown that the heuristics can provide useful information in defining the appropriate network architectur

    High performance associative memory models and weight dilution

    Get PDF
    The consequences of diluting the weights of the standard Hopfield architecture associative memory model, trained using perceptron like learning rules, is examined. A proportion of the weights of the network are removed; this can be done in a symmetric and asymmetric way and both methods are investigated. This paper reports experimental investigations into the consequences of dilution in terms of: capacity, training times and size of basins of attraction. It is concluded that these networks maintain a reasonable performance at fairly high dilution rates.Final Accepted Versio

    Global and Feature Based Gender Classification of Faces: A Comparison of Human Performance and Computational Models

    Get PDF
    Original paper can be found at: http://eproceedings.worldscinet.com/9789812701886/9789812701886_0036.html Copyright World Scientific Publishing Company. http://dx.doi.org/10.1142/9789812701886_0036Most computational models for gender classification use global information (the full face image) giving equal weight to the whole face area irrespective of the importance of the internal features. Here, we use a global and feature based representation of face images that includes both global and featural information. We use dimensionality reduction techniques and a support vector machine classifier and show that this method performs better than either global or feature based representations alone.Peer reviewe

    High Performance Associative Memories and Structured Weight Dilution

    Get PDF
    Copyright SpringerThe consequences of two techniques for symmetrically diluting the weights of the standard Hopfield architecture associative memory model, trained using a non-Hebbian learning rule, are examined. This paper reports experimental investigations into the effect of dilution on factors such as: pattern stability and attractor performance. It is concluded that these networks maintain a reasonable level of performance at fairly high dilution rates

    The analysis of animate object motion using neural networks and snakes

    Get PDF
    This paper presents a mechanism for analysing the deformable shape of an object as it moves across the visual field. An object’s outline is detected using active contour models, and is then re-represented as shape, location and rotation invariant axis crossover vectors. These vectors are used as input for a feedforward backpropagation neural network, which provides a confidence value determining how ‘human’ the network considers the given shape to be. The network was trained using simulated human shapes as well as simulated non-human shapes, including dogs, horses and inanimate objects. The network was then tested on unseen objects of these classes, as well as on an unseen object class. Analysis of the network’s confidence values for a given animated object identifies small, individual variations between different objects of the same class, and large variations between object classes. Confidence values for a given object are periodic and parallel the paces being taken by the object
    • 

    corecore